pascal voc
Supplementary Fairness Continual Learning Approach to Semantic Scene Understanding in Open-World Environments Thanh-Dat Truong
Contrastive Clustering loss and update the prototypical vectors.Algorithm 1: Prototypical Constrative Clustering Loss Compute Prototypical Constrative Clustering Loss based on Eqn. Compute Prototypical Constrative Clustering Loss based on Eqn. Two segmentation network architectures have been used in our experiments, i.e., (1) DeepLab-V3 The learning rate is set individually for each step and dataset. Similarly, to illustrate the effectiveness and robustness of our method in the non-incremental setting. We also perform an additional ablation study on the ADE20K (100-50) benchmark to investigate the impact of the delta.
- North America > United States > Arkansas > Washington County > Fayetteville (0.14)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- Workflow (0.49)
- Research Report > New Finding (0.35)
- Research Report > Strength High (1.00)
- Research Report > Experimental Study (1.00)
Selective Masking based Self-Supervised Learning for Image Semantic Segmentation
This paper proposes a novel self-supervised learning method for semantic segmentation using selective masking image reconstruction as the pretraining task. Our proposed method replaces the random masking augmentation used in most masked image modelling pretraining methods. The proposed selective masking method selectively masks image patches with the highest reconstruction loss by breaking the image reconstruction pretraining into iterative steps to leverage the trained model's knowledge. We show on two general datasets (Pascal VOC and Cityscapes) and two weed segmentation datasets (Nassar 2020 and Sugarbeets 2016) that our proposed selective masking method outperforms the traditional random masking method and supervised ImageNet pretraining on downstream segmentation accuracy by 2.9% for general datasets and 2.5% for weed segmentation datasets. Furthermore, we found that our selective masking method significantly improves accuracy for the lowest-performing classes. Lastly, we show that using the same pretraining and downstream dataset yields the best result for low-budget self-supervised pretraining. Our proposed Selective Masking Image Reconstruction method provides an effective and practical solution to improve end-to-end semantic segmentation workflows, especially for scenarios that require limited model capacity to meet inference speed and computational resource requirements.
- North America > Canada > Saskatchewan > Saskatoon (0.04)
- North America > United States > North Dakota > Williams County (0.04)
- Transportation > Ground (0.46)
- Health & Medicine > Diagnostic Medicine > Imaging (0.46)
- Information Technology > Sensing and Signal Processing > Image Processing (1.00)
- Information Technology > Artificial Intelligence > Vision (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.69)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.68)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.15)
- North America > United States > Arkansas > Washington County > Fayetteville (0.15)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.15)
Leveraging Out-of-Distribution Unlabeled Images: Semi-Supervised Semantic Segmentation with an Open-Vocabulary Model
Shin, Wooseok, Kang, Jisu, Jeong, Hyeonki, Kim, Jin Sob, Han, Sung Won
In semi-supervised semantic segmentation, existing studies have shown promising results in academic settings with controlled splits of benchmark datasets. However, the potential benefits of leveraging significantly larger sets of unlabeled images remain unexplored. In real-world scenarios, abundant unlabeled images are often available from online sources (web-scraped images) or large-scale datasets. However, these images may have different distributions from those of the target dataset, a situation known as out-of-distribution (OOD). Using these images as unlabeled data in semi-supervised learning can lead to inaccurate pseudo-labels, potentially misguiding network training. In this paper, we propose a new semi-supervised semantic segmentation framework with an open-vocabulary segmentation model (SemiOVS) to effectively utilize unlabeled OOD images. Extensive experiments on Pascal VOC and Context datasets demonstrate two key findings: (1) using additional unlabeled images improves the performance of semi-supervised learners in scenarios with few labels, and (2) using the open-vocabulary segmentation (OVS) model to pseudo-label OOD images leads to substantial performance gains. In particular, SemiOVS outperforms existing PrevMatch and SemiVL methods by +3.5 and +3.0 mIoU, respectively, on Pascal VOC with a 92-label setting, achieving state-of-the-art performance. These findings demonstrate that our approach effectively utilizes abundant unlabeled OOD images for semantic segmentation tasks. We hope this work can inspire future research and real-world applications. The code is available at https://github.com/wooseok-shin/SemiOVS
- North America > United States (0.14)
- Europe > Switzerland > Zürich > Zürich (0.14)
- Asia > South Korea > Seoul > Seoul (0.04)
- (2 more...)
- Education (0.67)
- Information Technology (0.67)
- Health & Medicine (0.46)
- Transportation > Ground (0.46)